Goto

Collaborating Authors

 reduction cell


Adapting Neural Architectures Between Domains (Supplementary Material) Y anxi Li1, Zhaohui Y ang 2,3, Yunhe Wang

Neural Information Processing Systems

By combining Theorem 2 and Lemma 3, we can derive the proof of Corollary 4. Let There are 2 kinds of cells in the search space, including normal cells and reduction cells. After a reduction cell, the channel number is doubled. Cells are stacked sequentially to build a network. We use a set of 8 different candidate operations, including: 3 3 separable convolution; 5 5 separable convolution; 3 3 dilated separable convolution; 5 5 dilated separable convolution; 3 3 max pooling; 3 3 average pooling; identity (i.e. All the operations follow the ReLU-Conv/Pooling-BN pattern except identity and zero.



Supplementary Materials for "ZARTS: On Zero-order Optimization for Neural Architecture Search " 1 Appendix 1.1 Estimation for Second-order Partial Derivative in DARTS

Neural Information Processing Systems

DARTS utilizes difference method, which is also a zero-order optimization algorithm.1.2 To draw loss landscapes w.r.t. In (b), we illustrate the landscape with second-order approximation. We fix iteration number M = 10 for all settings. Therefore, we remove zero operation from the search space. We apply Alg. 1 to train architecture parameters Models are trained for 600 epochs by SGD with a batch size of 96.


Supplementary Material of IST A-NAS: Efficient and Consistent Neural Architecture Search by Sparse Coding Yibo Y ang

Neural Information Processing Systems

We perform our experiments on both CIFAR-10 and ImageNet. The images are normalized by mean and standard deviation. The images are normalized by mean and standard deviation. Concretely, the super-net for search is composed of 6 normal cells and 2 reduction cells, and has an initial number of channels of 16. Each cell has 6 nodes.


Theory-Inspired Path-Regularized Differential Network Architecture Search (Supplementary File)

Neural Information Processing Systems

Then Appendix C gives the proofs of the main results in Sec. 3, namely Theorem 1, by first introducing auxiliary theories Due to space limitation, we defer more experimental results and details to this appendix. Due to the high training cost, we fix two regularization parameters and then investigate the third one. This testifies the robustness of PR-DARTS to regularization parameters.Figure 3: Effects of regularization parameters Here we first display the selected reduction cell on CIRAR10 in Figure 1 (a). Next, we also report the average gate activate probability in the normal and reduction cells in Figure 1 (b). At the beginning of the search, we initialize the activation probability of each gate to be one.


Supplementary Material of IST A-NAS: Efficient and Consistent Neural Architecture Search by Sparse Coding Yibo Y ang

Neural Information Processing Systems

We perform our experiments on both CIFAR-10 and ImageNet. The images are normalized by mean and standard deviation. The images are normalized by mean and standard deviation. Concretely, the super-net for search is composed of 6 normal cells and 2 reduction cells, and has an initial number of channels of 16. Each cell has 6 nodes.


Theory-Inspired Path-Regularized Differential Network Architecture Search (Supplementary File)

Neural Information Processing Systems

Then Appendix C gives the proofs of the main results in Sec. 3, namely Theorem 1, by first introducing auxiliary theories Due to space limitation, we defer more experimental results and details to this appendix. Due to the high training cost, we fix two regularization parameters and then investigate the third one. This testifies the robustness of PR-DARTS to regularization parameters.Figure 3: Effects of regularization parameters Here we first display the selected reduction cell on CIRAR10 in Figure 1 (a). Next, we also report the average gate activate probability in the normal and reduction cells in Figure 1 (b). At the beginning of the search, we initialize the activation probability of each gate to be one.